|
August 15, 2017
The Myth of General Purpose Wearables
August 8, 2017
IoT Connections Made Easy
August 1, 2017
Smarter Computing
July 25, 2017
The Value of Limits
July 18, 2017
Tech in the Heartland
June 27, 2017
Business Realities vs. Tech Dreams
June 20, 2017
The Power of Hidden Tech
June 13, 2017
Computing Evolves from Outside In to Inside Out
June 6, 2017
The Overlooked Surprises of Apple’s WWDC Keynote
May 30, 2017
Are AR and VR Only for Special Occasions?
May 23, 2017
The Digital Car
May 16, 2017
Digital Assistants Drive New Meta-Platform Battle
May 9, 2017
Getting Smart on Smart Speakers
May 5, 2017
Intel Opens High-Tech "Garage"
May 2, 2017
The Hidden Value of Analog
April 28, 2017
Google’s Waymo Starts Driving Passengers
April 25, 2017
The Robotic Future
April 21, 2017
Sony Debuts New Pro Camera
April 18, 2017
Should Apple Build a Car?
April 14, 2017
PC Market Outlook Improving
April 11, 2017
Little Data Analytics
April 7, 2017
Facebook Debuts Free Version of Workplace Collaboration Tool
April 4, 2017
Samsung Building a Platform Without an OS
March 31, 2017
Microsoft Announces Windows 10 Creators Update Release Date
March 28, 2017
Augmented Reality Finally Delivers on 3D Promise
March 24, 2017
Intel Creates AI Organization
March 21, 2017
Chip Magic
March 17, 2017
Microsoft Unveils Teams Chat App
March 14, 2017
Computing on the Edge
March 7, 2017
Cars Need Digital Safety Standards Too
February 28, 2017
The Messy Path to 5G
February 24, 2017
AMD Launches Ryzen CPU
February 21, 2017
Rethinking Wearable Computing
February 17, 2017
Samsung Heir Arrest Unlikely to Impact Sales
February 14, 2017
Modern Workplaces Still More Vision Than Reality
February 10, 2017
Lenovo Develops Energy-Efficient Soldering Technology
February 7, 2017
The Missing Map from Silicon Valley to Main Street
January 31, 2017
The Network vs. The Computer
January 27, 2017
Facebook Adds Support For FIDO Security Keys
January 24, 2017
Voice Drives New Software Paradigm
January 20, 2017
Tesla Cleared of Fault in NHTSA Crash Probe
January 17, 2017
Inside the Mind of a Hacker
January 13, 2017
PC Shipments Stumble but Turnaround is Closer
January 10, 2017
Takeaways from CES 2017
January 3, 2017
Top 10 Tech Predictions for 2017
|
|
|
|
August 22, 2017
By Bob O'Donnell
For a relatively nascent product category, smart speakers like Amazon Echo and Google Home are already seeing a huge influx of attention from both consumers and potential competitors eager to enter the market. Apple has announced the HomePod and numerous other vendors have either unveiled or are heavily rumored to be working on versions of their own.
Harman Kardon (in conjunction with Microsoft), GE Lighting and Lenovo have announced products in the US, while Alibaba, Xiaomi and JD.com, among others, have said they will be bringing products out in China. In addition, Facebook is rumored to be building a screen-equipped smart speaker called Gizmo.
One obvious question after hearing about all the new entrants is, how can they all survive? The short answer, of course, is they won’t. Nevertheless, expect to see a lot of jockeying, marketing and positioning over the next year or two because it’s still very early days in the world of AI-powered and personal assistant-driven smart speakers.
Yes, Amazon has built an impressive and commanding presence with the Echo line, but there are many limitations to Echos and all current smart speakers that frustrate existing users. Thankfully, technology improvements are coming that will enable competitors to differentiate themselves from others in ways which reduce the frustration and increase the satisfaction that consumers have with smart speakers.
Part of the work involves the overall architecture of the devices and how they interact with cloud-based services. For example, one of the critical capabilities that many users want is the ability to accurately recognize different individuals that speak to the device, so that responses can be customized for different members of a household. To achieve this as quickly and accurately as possible, it doesn’t make sense to try and send the audio signal to the cloud and then wait for the response. Even with superfast network connections, the inevitable delays make interactions with the device feel somewhat awkward.
The same problem exists when you try to move beyond the simple single query requests that most people are making to their smart speakers today. (Alexa, play music by horn bands or Alexa, what is the capital of Iceland?) In order to have naturally flowing, multi-question or multi-statement conversations, the delays (or latency) have to be dramatically reduced.
The obvious answer to the problem is to do more of the recognition and response work locally on the device and not rely on a cloud-based network connection to do so. In fact, this is a great example of the larger trend of edge computing, where we are seeing devices or applications that use to rely solely on big data centers in the cloud start to do more of the computational work on their own.
That’s part of the reason you’re starting to see companies like Qualcomm and Intel, among others, develop chips that are designed to enable more powerful local computing work on devices like smart speakers. The ability to learn and then recognize different individuals, for example, is something that the DSP (digital signal processor) component of new chips from these vendors can do.
Another technological challenge facing current generation products is recognition accuracy. Everyone who has used a smart speaker or digital assistant on other device has had the experience of not being understood. Sometimes that’s due to how the question or command is phrased, but it’s often due to background noises, accents, intonation or other factors that essentially end up providing an imperfect audio signal to the cloud-based recognition engine. Again, more local audio signal processing can often improve the audio signal to be sent, thereby enhancing overall recognition.
Going further, most of the AI-based learning algorithms used to recognize and accurately respond to speech will likely need to be run in very large, compute-intensive cloud data centers. However, the idea of being able to start do pattern recognition of common phrases (a form of inferencing—the second key aspect of machine learning and AI) locally with the right kind of computing engines and hardware architectures is becoming increasingly possible. It may be a long time before all that kind of work can be done within smart speakers and other edge devices, but even doing some speech recognition on the device should enable higher accuracy and longer conversations. In short, a much better user experience.
As new entrants try to differentiate their products in an increasingly crowded space, the ability to offer some key tech-based improvements is going to be essential. Clearly there’s a great deal of momentum behind the smart speaker phenomenon, but it’s going to take these kind performance improvements to move them beyond idle curiosities and into truly useful, everyday kinds of tools.
Here's a link to the column: https://techpinions.com/the-evolution-of-smart-speakers/50864
Bob O’Donnell is the president and chief analyst of TECHnalysis Research, LLC a market research firm that provides strategic consulting and market research services to the technology industry and professional financial community. You can follow him on Twitter @bobodtech.
Podcasts
Leveraging more than 10 years of award-winning, professional radio experience, TECHnalysis Research participates in a video-based podcast called Everything Technology.
LEARN MORE |
|
Research Offerings
TECHnalysis Research offers a wide range of research deliverables that you can read about here.
READ MORE |
|